# Long-text understanding

Qwen3 4B Base
Apache-2.0
Qwen3-4B-Base is the latest generation of the Qwen series' 4-billion-parameter large language model, pre-trained on 36 trillion tokens of multilingual data, supporting a 32k context length.
Large Language Model Transformers
Q
Qwen
50.84k
29
Ultralong Thinking
An 8B-parameter language model merged using the SLERP method, combining the strengths of DeepSeek-R1 and Nemotron-8B models
Large Language Model Transformers
U
mergekit-community
69
2
Modernbert Large Nli
Apache-2.0
A multi-task fine-tuned model based on ModernBERT-large, specializing in Natural Language Inference (NLI) tasks, excelling in zero-shot classification and reasoning tasks.
Large Language Model Transformers Supports Multiple Languages
M
p-christ
39
0
LLM2CLIP Openai B 16
Apache-2.0
LLM2CLIP is an innovative method that leverages large language models (LLMs) to extend CLIP's capabilities, enhancing text discriminability through a contrastive learning framework and significantly improving cross-modal task performance.
Text-to-Image Safetensors
L
microsoft
1,154
18
LLM2CLIP EVA02 L 14 336
Apache-2.0
LLM2CLIP is an innovative approach that enhances CLIP's visual representation capabilities through large language models (LLMs), significantly improving cross-modal task performance
Text-to-Image PyTorch
L
microsoft
75
60
Llama3 8B 1.58 100B Tokens
Large language model fine-tuned based on BitNet 1.58b architecture, with Llama-3-8B-Instruct as the base model, employing extreme quantization techniques
Large Language Model Transformers
L
HF1BitLLM
2,427
181
Tess V2.5 Phi 3 Medium 128k 14B
MIT
A large language model fine-tuned based on Microsoft Phi-3-medium-128k-instruct, supporting ChatML format dialogue interaction
Large Language Model Transformers
T
migtissera
4,932
4
Yi 1.5 6B Chat
Apache-2.0
Yi-1.5 is an upgraded version of the Yi model, excelling in programming, mathematics, reasoning, and instruction-following capabilities while maintaining outstanding language understanding, commonsense reasoning, and reading comprehension abilities.
Large Language Model Transformers
Y
01-ai
13.32k
42
Mistral 7B V0.1 Flashback V2
MIT
A pre-trained continuation model based on Mistral-7B-v0.1, fine-tuned with 40GB of text data from the Swedish forum Flashback, supporting multilingual generation.
Large Language Model Transformers Supports Multiple Languages
M
timpal0l
98
9
Bloomz 3b Nli
Openrail
A natural language inference model fine-tuned based on Bloomz-3b-chat-dpo, supporting semantic relation judgment in English and French
Large Language Model Transformers Supports Multiple Languages
B
cmarkea
22
1
Xlm Roberta Large Squad2 Qa Milqa Impossible
This model is a Hungarian question-answering model fine-tuned on the milqa dataset based on deepset/xlm-roberta-large-squad2, supporting handling cases with no answers.
Question Answering System Transformers Other
X
ZTamas
28
2
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase